Training set approximation for kernel methods

نویسندگان

  • Vojtěch Franc
  • Václav Hlaváč
چکیده

We propose a technique for a training set approximation and its usage in kernel methods. The approach aims to represent data in a low dimensional space with possibly minimal representation error which is similar to the Principal Component Analysis (PCA). In contrast to the PCA, the basis vectors of the low dimensional space used for data representation are properly selected vectors from the training set and not as their linear combinations. The basis vectors can be selected by a simple algorithm which has low computational requirements and allows on-line processing of huge data sets. The proposed method was used to approximate training sets of the Support Vector Machines and Kernel Fisher Linear Discriminant which are known method for learning classifiers. The experiments show that the proposed approximation can significantly reduce the complexity of the found classifiers (the number of the support vectors) while retaining their accuracy.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Greedy Algorithm for a Training Set Reduction in the Kernel Methods

We propose a technique for a training set approximation and its usage in kernel methods. The approach aims to represent data in a low dimensional space with possibly minimal representation error which is similar to the Principal Component Analysis (PCA). In contrast to the PCA, the basis vectors of the low dimensional space used for data representation are properly selected vectors from the tra...

متن کامل

Scalable Kernel Systems

Kernel-based systems are currently very popular approaches to supervised learning. Unfortunately, the computational load for training kernel-based systems increases drastically with the number of training data points. Recently, a number of approximate methods for scaling kernel-based systems to large data sets have been introduced. In this paper we investigate the relationship between three of ...

متن کامل

Large-Scale Support Vector Machine Training using Low-Dimensional Approximation of the Kernel Matrix

Training Support Vector Machines (regression and/or classification) involves solving a simply constrained quadratic programming problem that scales linearly with training set size. Unfortunately this makes solving large-scale problems with millions of training samples difficult if not impossible from both computational and storage considerations. A standard approach is to exploit the sparsity o...

متن کامل

On Approximate Solutions to Support Vector Machines

We propose to speed up the training process of support vector machines (SVM) by resorting to an approximate SVM, where a small number of representatives are extracted from the original training data set and used for training. Theoretical studies show that, in order for the approximate SVM to be similar to the exact SVM given by the original training data set, kernel k-means should be used to ex...

متن کامل

A robust autoregressive gaussian process motion model using l1-norm based low-rank kernel matrix approximation

This paper considers the problem of modeling complex motions of pedestrians in a crowded environment. A number of methods have been proposed to predict the motion of a pedestrian or an object. However, it is still difficult to make a good prediction due to challenges, such as the complexity of pedestrian motions and outliers in a training set. This paper addresses these issues by proposing a ro...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2003